多药物(定义为使用多种药物)是一种标准治疗方法,尤其是对于严重和慢性疾病。但是,将多种药物一起使用可能会导致药物之间的相互作用。药物 - 药物相互作用(DDI)是一种与另一种药物结合时的影响发生变化时发生的活性。 DDI可能会阻塞,增加或减少药物的预期作用,或者在最坏情况下,会产生不利的副作用。虽然准时检测DDI至关重要,但由于持续时间短,并且在临床试验中识别它们是时间的,而且昂贵,并且要考虑许多可能的药物对进行测试。结果,需要计算方法来预测DDI。在本文中,我们提出了一种新型的异质图注意模型Han-DDI,以预测药物 - 药物相互作用。我们建立了具有不同生物实体的药物网络。然后,我们开发了一个异质的图形注意网络,以使用药物与其他实体的关系学习DDI。它由一个基于注意力的异质图节点编码器组成,用于获得药物节点表示和用于预测药物相互作用的解码器。此外,我们利用全面的实验来评估我们的模型并将其与最先进的模型进行比较。实验结果表明,我们提出的方法Han-DDI的表现可以显着,准确地预测DDI,即使对于新药也是如此。
translated by 谷歌翻译
药物 - 药物相互作用(DDIS)可能会阻碍药物的功能,在最坏的情况下,它们可能导致不良药物反应(ADR)。预测所有DDI是一个具有挑战性且关键的问题。大多数现有的计算模型都集成了来自不同来源的药物中心信息,并利用它们作为机器学习分类器中的功能来预测DDIS。但是,这些模型有很大的失败机会,尤其是对于所有信息都没有可用的新药。本文提出了一个新型的HyperGraph神经网络(HYGNN)模型,仅基于用于DDI预测问题的任何药物的微笑串。为了捕获药物的相似性,我们创建了从微笑字符串中提取的药物的化学子结构中创建的超图。然后,我们开发了由新型的基于注意力的超图边缘编码器组成的HYGNN,以使药物的表示形式和解码器,以预测药物对之间的相互作用。此外,我们进行了广泛的实验,以评估我们的模型并将其与几种最新方法进行比较。实验结果表明,我们提出的HYGNN模型有效地预测了DDI,并以最大的ROC-AUC和PR-AUC分别超过基准,分别为97.9%和98.1%。
translated by 谷歌翻译
肝脏是脊椎动物中最关键的代谢器官之一,由于其在人体中的重要功能,例如废物产物和药物的血液排毒。由于肝肿瘤引起的肝病是全球最常见的死亡率之一。因此,在肿瘤发育的早期阶段检测肝肿瘤是医疗治疗的关键部分。许多成像方式可以用作检测肝肿瘤的帮助工具。计算机断层扫描(CT)是软组织器官(例如肝脏)最常用的成像方式。这是因为它是一种侵入性方式,可以相对迅速捕获。本文提出了一个有效的自动肝分割框架,以使用3D CNN深度元网络模型检测和分割肝脏腹部扫描。许多研究采用了精确分割肝区域,然后使用分割的肝区域作为肿瘤分割方法的输入,因为它降低了由于将腹部器官分割为肿瘤而导致的错误率。所提出的3D CNN DeepMedic模型具有两个输入途径,而不是一个途径,如原始3D CNN模型所示。在本文中,该网络提供了多个腹部CT版本,这有助于提高细分质量。提出的模型分别达到94.36%,94.57%,91.86%和93.14%的精度,灵敏度,特异性和骰子相似性得分。实验结果表明该方法的适用性。
translated by 谷歌翻译
开发旨在增强胎儿监测的创新信息学方法是生殖医学研究的新领域。已经对人工智能(AI)技术进行了几项评论,以改善妊娠结局。他们的限制是专注于特定数据,例如怀孕期间母亲的护理。这项系统的调查旨在探讨人工智能(AI)如何通过超声(US)图像帮助胎儿生长监测。我们使用了八个医学和计算机科学书目数据库,包括PubMed,Embase,Psycinfo,ScienceDirect,IEEE Explore,ACM图书馆,Google Scholar和Web of Science。我们检索了2010年至2021年之间发表的研究。从研究中提取的数据是使用叙述方法合成的。在1269项检索研究中,我们包括了107项与调查中有关该主题的查询的不同研究。我们发现,与3D和4D超声图像(n = 19)相比,2D超声图像更受欢迎(n = 88)。分类是最常用的方法(n = 42),其次是分割(n = 31),与分割(n = 16)集成的分类和其他其他杂项,例如对象检测,回归和增强学习(n = 18)。妊娠结构域中最常见的区域是胎儿头(n = 43),然后是胎儿(n = 31),胎儿心脏(n = 13),胎儿腹部(n = 10),最后是胎儿的面孔(n = 10)。在最近的研究中,深度学习技术主要使用(n = 81),其次是机器学习(n = 16),人工神经网络(n = 7)和增强学习(n = 2)。 AI技术在预测胎儿疾病和鉴定怀孕期间胎儿解剖结构中起着至关重要的作用。需要进行更多的研究来从医生的角度验证这项技术,例如试点研究和有关AI及其在医院环境中的应用的随机对照试验。
translated by 谷歌翻译
Artificial Intelligence (AI) has become commonplace to solve routine everyday tasks. Because of the exponential growth in medical imaging data volume and complexity, the workload on radiologists is steadily increasing. We project that the gap between the number of imaging exams and the number of expert radiologist readers required to cover this increase will continue to expand, consequently introducing a demand for AI-based tools that improve the efficiency with which radiologists can comfortably interpret these exams. AI has been shown to improve efficiency in medical-image generation, processing, and interpretation, and a variety of such AI models have been developed across research labs worldwide. However, very few of these, if any, find their way into routine clinical use, a discrepancy that reflects the divide between AI research and successful AI translation. To address the barrier to clinical deployment, we have formed MONAI Consortium, an open-source community which is building standards for AI deployment in healthcare institutions, and developing tools and infrastructure to facilitate their implementation. This report represents several years of weekly discussions and hands-on problem solving experience by groups of industry experts and clinicians in the MONAI Consortium. We identify barriers between AI-model development in research labs and subsequent clinical deployment and propose solutions. Our report provides guidance on processes which take an imaging AI model from development to clinical implementation in a healthcare institution. We discuss various AI integration points in a clinical Radiology workflow. We also present a taxonomy of Radiology AI use-cases. Through this report, we intend to educate the stakeholders in healthcare and AI (AI researchers, radiologists, imaging informaticists, and regulators) about cross-disciplinary challenges and possible solutions.
translated by 谷歌翻译
State space models (SSMs) have demonstrated state-of-the-art sequence modeling performance in some modalities, but underperform attention in language modeling. Moreover, despite scaling nearly linearly in sequence length instead of quadratically, SSMs are still slower than Transformers due to poor hardware utilization. In this paper, we make progress on understanding the expressivity gap between SSMs and attention in language modeling, and on reducing the hardware barrier between SSMs and attention. First, we use synthetic language modeling tasks to understand the gap between SSMs and attention. We find that existing SSMs struggle with two capabilities: recalling earlier tokens in the sequence and comparing tokens across the sequence. To understand the impact on language modeling, we propose a new SSM layer, H3, that is explicitly designed for these abilities. H3 matches attention on the synthetic languages and comes within 0.4 PPL of Transformers on OpenWebText. Furthermore, a hybrid 125M-parameter H3-attention model that retains two attention layers surprisingly outperforms Transformers on OpenWebText by 1.0 PPL. Next, to improve the efficiency of training SSMs on modern hardware, we propose FlashConv. FlashConv uses a fused block FFT algorithm to improve efficiency on sequences up to 8K, and introduces a novel state passing algorithm that exploits the recurrent properties of SSMs to scale to longer sequences. FlashConv yields 2$\times$ speedup on the long-range arena benchmark and allows hybrid language models to generate text 1.6$\times$ faster than Transformers. Using FlashConv, we scale hybrid H3-attention language models up to 1.3B parameters on the Pile and find promising initial results, achieving lower perplexity than Transformers and outperforming Transformers in zero- and few-shot learning on a majority of tasks in the SuperGLUE benchmark.
translated by 谷歌翻译
Increasing popularity of deep-learning-powered applications raises the issue of vulnerability of neural networks to adversarial attacks. In other words, hardly perceptible changes in input data lead to the output error in neural network hindering their utilization in applications that involve decisions with security risks. A number of previous works have already thoroughly evaluated the most commonly used configuration - Convolutional Neural Networks (CNNs) against different types of adversarial attacks. Moreover, recent works demonstrated transferability of the some adversarial examples across different neural network models. This paper studied robustness of the new emerging models such as SpinalNet-based neural networks and Compact Convolutional Transformers (CCT) on image classification problem of CIFAR-10 dataset. Each architecture was tested against four White-box attacks and three Black-box attacks. Unlike VGG and SpinalNet models, attention-based CCT configuration demonstrated large span between strong robustness and vulnerability to adversarial examples. Eventually, the study of transferability between VGG, VGG-inspired SpinalNet and pretrained CCT 7/3x1 models was conducted. It was shown that despite high effectiveness of the attack on the certain individual model, this does not guarantee the transferability to other models.
translated by 谷歌翻译
Can a neural network estimate an object's dimension in the wild? In this paper, we propose a method and deep learning architecture to estimate the dimensions of a quadrilateral object of interest in videos using a monocular camera. The proposed technique does not use camera calibration or handcrafted geometric features; however, features are learned with the help of coefficients of a segmentation neural network during the training process. A real-time instance segmentation-based Deep Neural Network with a ResNet50 backbone is employed, giving the object's prototype mask and thus provides a region of interest to regress its dimensions. The instance segmentation network is trained to look at only the nearest object of interest. The regression is performed using an MLP head which looks only at the mask coefficients of the bounding box detector head and the prototype segmentation mask. We trained the system with three different random cameras achieving 22% MAPE for the test dataset for the dimension estimation
translated by 谷歌翻译
In recent years, the exponential proliferation of smart devices with their intelligent applications poses severe challenges on conventional cellular networks. Such challenges can be potentially overcome by integrating communication, computing, caching, and control (i4C) technologies. In this survey, we first give a snapshot of different aspects of the i4C, comprising background, motivation, leading technological enablers, potential applications, and use cases. Next, we describe different models of communication, computing, caching, and control (4C) to lay the foundation of the integration approach. We review current state-of-the-art research efforts related to the i4C, focusing on recent trends of both conventional and artificial intelligence (AI)-based integration approaches. We also highlight the need for intelligence in resources integration. Then, we discuss integration of sensing and communication (ISAC) and classify the integration approaches into various classes. Finally, we propose open challenges and present future research directions for beyond 5G networks, such as 6G.
translated by 谷歌翻译
Pruning refers to the elimination of trivial weights from neural networks. The sub-networks within an overparameterized model produced after pruning are often called Lottery tickets. This research aims to generate winning lottery tickets from a set of lottery tickets that can achieve similar accuracy to the original unpruned network. We introduce a novel winning ticket called Cyclic Overlapping Lottery Ticket (COLT) by data splitting and cyclic retraining of the pruned network from scratch. We apply a cyclic pruning algorithm that keeps only the overlapping weights of different pruned models trained on different data segments. Our results demonstrate that COLT can achieve similar accuracies (obtained by the unpruned model) while maintaining high sparsities. We show that the accuracy of COLT is on par with the winning tickets of Lottery Ticket Hypothesis (LTH) and, at times, is better. Moreover, COLTs can be generated using fewer iterations than tickets generated by the popular Iterative Magnitude Pruning (IMP) method. In addition, we also notice COLTs generated on large datasets can be transferred to small ones without compromising performance, demonstrating its generalizing capability. We conduct all our experiments on Cifar-10, Cifar-100 & TinyImageNet datasets and report superior performance than the state-of-the-art methods.
translated by 谷歌翻译